Fully Variational Noise-Contrastive Estimation
نویسندگان
چکیده
By using the underlying theory of proper scoring rules, we design a family noise-contrastive estimation (NCE) methods that are tractable for latent variable models. Both terms in NCE loss, one data samples and noise samples, can be lower-bounded as variational Bayes, therefore call this losses fully estimation. Variational autoencoders particular example also understood separating real from synthetic an appropriate classification loss. We further discuss other instances objectives indicate differences their empirical behavior.
منابع مشابه
Improving Language Modelling with Noise-contrastive estimation
Neural language models do not scale well when the vocabulary is large. Noise contrastive estimation (NCE) is a sampling-based method that allows for fast learning with large vocabularies. Although NCE has shown promising performance in neural machine translation, its full potential has not been demonstrated in the language modelling literature. A sufficient investigation of the hyperparameters ...
متن کاملLearning Probabilistic Submodular Diversity Models Via Noise Contrastive Estimation
Modeling diversity of sets of items is important in many applications such as product recommendation and data summarization. Probabilistic submodular models, a family of models including the determinantal point process, form a natural class of distributions, encouraging effects such as diversity, repulsion and coverage. Current models, however, are limited to small and medium number of items du...
متن کاملNotes on Noise Contrastive Estimation and Negative Sampling
Estimating the parameters of probabilistic models of language such as maxent models and probabilistic neural models is computationally difficult since it involves evaluating partition functions by summing over an entire vocabulary, which may be millions of word types in size. Two closely related strategies—noise contrastive estimation (Mnih and Teh, 2012; Mnih and Kavukcuoglu, 2013; Vaswani et ...
متن کاملCollective Noise Contrastive Estimation for Policy Transfer Learning
We address the problem of learning behaviour policies to optimise online metrics from heterogeneous usage data. While online metrics, e.g., click-through rate, can be optimised effectively using exploration data, such data is costly to collect in practice, as it temporarily degrades the user experience. Leveraging related data sources to improve online performance would be extremely valuable, b...
متن کاملLearning word embeddings efficiently with noise-contrastive estimation
Continuous-valued word embeddings learned by neural language models have recently been shown to capture semantic and syntactic information about words very well, setting performance records on several word similarity tasks. The best results are obtained by learning high-dimensional embeddings from very large quantities of data, which makes scalability of the training method a critical factor. W...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Lecture Notes in Computer Science
سال: 2023
ISSN: ['1611-3349', '0302-9743']
DOI: https://doi.org/10.1007/978-3-031-31438-4_12